-
Notifications
You must be signed in to change notification settings - Fork 234
feat: Make supported hadoop filesystem schemes configurable #2272
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Codecov Report❌ Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## main #2272 +/- ##
============================================
+ Coverage 56.12% 57.90% +1.77%
- Complexity 976 1291 +315
============================================
Files 119 146 +27
Lines 11743 13376 +1633
Branches 2251 2374 +123
============================================
+ Hits 6591 7745 +1154
- Misses 4012 4374 +362
- Partials 1140 1257 +117 ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
} | ||
.map_err(|e| ExecutionError::GeneralError(e.to_string()))?; | ||
let (object_store, object_store_path): (Box<dyn ObjectStore>, Path) = | ||
if is_hdfs_scheme(&url, object_store_configs) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There is a little gotcha here when the scheme is s3a. In s3a's case, we replace s3a with s3 so that we can use the native object store implementation (
if scheme == "s3a" { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks, I made some adjustments, could you take another look?
@@ -40,6 +41,8 @@ object NativeConfig { | |||
// Azure Data Lake Storage Gen2 secure configurations (can use both prefixes) | |||
"abfss" -> Seq("fs.abfss.", "fs.abfs.")) | |||
|
|||
val COMET_LIBHDFS_SCHEMES_KEY = "fs.comet.libhdfs.schemes" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should we make this a Comet conf (i.e add it in CometConf
so it is automatically documented)?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks, moved to CometConf.
conf(s"spark.hadoop.$COMET_LIBHDFS_SCHEMES_KEY") | ||
.doc( | ||
"Defines filesystem schemes (e.g., hdfs, webhdfs) that the native side accesses " + | ||
"via libhdfs, separated by commas.") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nit: perhaps we can mention that this configuration is valid only if comet has been built with the hdfs feature flag enabled.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks, added this description
Which issue does this PR close?
Closes #2271.
Rationale for this change
Currently we prefer to use jvm-based libhdfs to implement native hdfs reader, which means we can support more hadoop file systems. But currently we hardcode to support only hdfs scheme, I want to make the supported hadoop file system schemes configurable.
What changes are included in this PR?
Make supported hadoop filesystem schemes configurable
How are these changes tested?
After patch #2244, the newly added test cases were successfully run